3 research outputs found

    Design and implementation of a cloud computing service for finite element analysis

    Get PDF
    This paper presents an end-to-end discussion on the technical issues related to the design and implementation of a new cloud computing service for finite element analysis (FEA). The focus is specifically on performance characterization of linear and nonlinear mechanical structural analysis workloads over multi-core and multi-node computing resources. We first analyze and observe that accurate job characterization, tuning of multi-threading parameters and effective multi-core/node scheduling are critical for service performance. We design a “smart” scheduler that can dynamically select some of the required parameters, partition the load and schedule it in a resource-aware manner. We can achieve up to 7.53× performance improvement over an aggressive scheduler using mixed FEA loads. We also discuss critical issues related to the data privacy, security, accounting, and portability of the cloud service.European Commission ; IBM Shared University Research (SUR) program ; TÜBİTAK ; IBM PhD Fellowship awardpost-prin

    Smart job scheduling for high-performance cloud computing services

    No full text
    Due to copyright restrictions, the access to the full text of this article is only available via subscription.In this paper, we describe the challenges faced and lessons learned while establishing a large-scale high performance cloud computing service that enables online mechanical structural analysis and many other scientific applications using the finite element analysis (FEA) technique. The service is intended to process many independent and loosely-dependent (e.g. assembled system) tasks concurrently. Challenges faced include accurate job characterization, handling of many-task mixed jobs, sensitivity of task execution to multi-threading parameters, effective multi-core scheduling in a single node, and achieving seamless scale across multiple nodes. We find that significant performance gains in terms of both job completion latency and throughput are possible via dynamic or "smart" partitioning and resource-aware scheduling compared to shortest first and aggressive job scheduling techniques. We also discuss issues related to secure and private processing of sensitive models in the cloud

    Democratization of HPC cloud services with automated parallel solvers and application containers

    No full text
    Due to copyright restrictions, the access to the full text of this article is only available via subscription.In this paper, we investigate several design choices for HPC services at different layers of the cloud computing architecture to simplify and broaden its use cases. We start with the platform-as-a-service (PaaS) layer and compare direct and iterative parallel linear equation solvers. We observe that several matrix properties that can be identified before starting long-running solvers can help HPC services automatically select the amount of computing resources per job, such that the job latency is minimized and the overall job throughput is maximized. As a proof of concept, we use classical problems in structural mechanics and mesh these problems with increasing granularities leading to various matrix sizes, ie, largest having 1 billion non-zero elements. In addition to matrix size, we take into account matrix condition numbers, preconditioning effects, and solver types and execute these finite element analysis (FEA) over an IBM HPC cluster. Next, we focus on the infrastructure-as-a-service (IaaS) layer and explore HPC application performance, load isolation, and deployment issues using application containers (Docker) while also comparing them to physical and virtual machines (VM) over a public cloud.IBM Faculty Award ; IBM PhD Fellowship programs ; EU Marie Curie FP7 BI4MASSES projec
    corecore